Goto

Collaborating Authors

 string space



BOSS: Bayesian Optimization over String Spaces

Neural Information Processing Systems

This article develops a Bayesian optimization (BO) method which acts directly over raw strings, proposing the first uses of string kernels and genetic algorithms within BO loops. Recent applications of BO over strings have been hindered by the need to map inputs into a smooth and unconstrained latent space. Learning this projection is computationally and data-intensive. Our approach instead builds a powerful Gaussian process surrogate model based on string kernels, naturally supporting variable length inputs, and performs efficient acquisition function maximization for spaces with syntactic constraints. Experiments demonstrate considerably improved optimization over existing approaches across a broad range of constraints, including the popular setting where syntax is governed by a context-free grammar.



contribution to the NeurIPS community, providing a needed solution for a problem with immediate applications across 3

Neural Information Processing Systems

In this rebuttal, we respond to each reviewer's minor comments individually, as there are no Thank you for the extremely positive review. We will fix the one typo reported. Rubanova et al. [2020] also consider Bayesian optimisation over strings. We will discuss this contemporaneous work (published at the start of this month) in our final version.


Cross-Entropy Games for Language Models: From Implicit Knowledge to General Capability Measures

Hongler, Clément, Emil, Andrew

arXiv.org Artificial Intelligence

Large Language Models (LLMs) define probability measures on text. By considering the implicit knowledge question of what it means for an LLM to know such a measure and what it entails algorithmically, we are naturally led to formulate a series of tasks that go beyond generative sampling, involving forms of summarization, counterfactual thinking, anomaly detection, originality search, reverse prompting, debating, creative solving, etc. These tasks can be formulated as games based on LLM measures, which we call Cross-Entropy (Xent) Games . Xent Games can be single-player or multi-player. They involve cross-entropy scores and cross-entropy constraints, and can be expressed as simple computational graphs and programs. We show the Xent Game space is large enough to contain a wealth of interesting examples, while being constructible from basic game-theoretic consistency axioms. We then discuss how the Xent Game space can be used to measure the abilities of LLMs. This leads to the construction of Xent Game measures: finite families of Xent Games that can be used as capability benchmarks, built from a given scope, by extracting a covering measure. To address the unbounded scope problem associated with the challenge of measuring general abilities, we propose to explore the space of Xent Games in a coherent fashion, using ideas inspired by evolutionary dynamics.


Review for NeurIPS paper: BOSS: Bayesian Optimization over String Spaces

Neural Information Processing Systems

Additional Feedback: (1) I am curious about the differences between the proposed work and this one: Amortized Bayesian Optimization over Discrete Spaces, by Yulia Rubanov, etc. It seems the general idea is quite similar, for example, they also employ the evolutionary algorithms in BO. (2) What's the maximized length of the string did you test in the experiment? Besides the GA, I am curious if the authors have tried some other methods, such as random forest. In your scenario, have you ever considered using different kernels to build the objective function? Obviously, this work will be welcomed in the community. By considering the original good work and the satisfied feedback, I'd like to rise my score to "A good submission, accept".


BOSS: Bayesian Optimization over String Spaces

Neural Information Processing Systems

This article develops a Bayesian optimization (BO) method which acts directly over raw strings, proposing the first uses of string kernels and genetic algorithms within BO loops. Recent applications of BO over strings have been hindered by the need to map inputs into a smooth and unconstrained latent space. Learning this projection is computationally and data-intensive. Our approach instead builds a powerful Gaussian process surrogate model based on string kernels, naturally supporting variable length inputs, and performs efficient acquisition function maximization for spaces with syntactic constraints. Experiments demonstrate considerably improved optimization over existing approaches across a broad range of constraints, including the popular setting where syntax is governed by a context-free grammar.


BOSS: Bayesian Optimization over String Spaces

Moss, Henry B., Beck, Daniel, Gonzalez, Javier, Leslie, David S., Rayson, Paul

arXiv.org Artificial Intelligence

This article develops a Bayesian optimization (BO) method which acts directly over raw strings, proposing the first uses of string kernels and genetic algorithms within BO loops. Recent applications of BO over strings have been hindered by the need to map inputs into a smooth and unconstrained latent space. Learning this projection is computationally and data-intensive. Our approach instead builds a powerful Gaussian process surrogate model based on string kernels, naturally supporting variable length inputs, and performs efficient acquisition function maximization for spaces with syntactical constraints. Experiments demonstrate considerably improved optimization over existing approaches across a broad range of constraints, including the popular setting where syntax is governed by a context-free grammar.